Goto

Collaborating Authors

 generic optimization



GENO -- GENeric Optimization for Classical Machine Learning

Neural Information Processing Systems

Although optimization is the longstanding, algorithmic backbone of machine learning new models still require the time-consuming implementation of new solvers. As a result, there are thousands of implementations of optimization algorithms for machine learning problems. A natural question is, if it is always necessary to implement a new solver, or is there one algorithm that is sufficient for most models. Common belief suggests that such a one-algorithm-fits-all approach cannot work, because this algorithm cannot exploit model specific structure. At least, a generic algorithm cannot be efficient and robust on a wide variety of problems.


GENO -- GENeric Optimization for Classical Machine Learning

Soeren Laue, Matthias Mitterreiter, Joachim Giesen

Neural Information Processing Systems

Although optimization is the longstanding algorithmic backbone of machine learning, new models still require the time-consuming implementation of new solvers. As a result, there are thousands of implementations of optimization algorithms for machine learning problems.



Reviews: GENO -- GENeric Optimization for Classical Machine Learning

Neural Information Processing Systems

The paper presents a new software framework for automatic generation of efficient solvers for a variety of optimization problems. Reviewers uniformly liked the generic approach and the use of automatic differentiation on a symbolic level. Based on the consensus, the paper is accepted, and we hope the authors will implement the suggestions provided in the reviews.


GENO -- GENeric Optimization for Classical Machine Learning

Neural Information Processing Systems

Although optimization is the longstanding, algorithmic backbone of machine learning new models still require the time-consuming implementation of new solvers. As a result, there are thousands of implementations of optimization algorithms for machine learning problems. A natural question is, if it is always necessary to implement a new solver, or is there one algorithm that is sufficient for most models. Common belief suggests that such a one-algorithm-fits-all approach cannot work, because this algorithm cannot exploit model specific structure. At least, a generic algorithm cannot be efficient and robust on a wide variety of problems.


GENO -- GENeric Optimization for Classical Machine Learning

Laue, Soeren, Mitterreiter, Matthias, Giesen, Joachim

Neural Information Processing Systems

Although optimization is the longstanding, algorithmic backbone of machine learning new models still require the time-consuming implementation of new solvers. As a result, there are thousands of implementations of optimization algorithms for machine learning problems. A natural question is, if it is always necessary to implement a new solver, or is there one algorithm that is sufficient for most models. Common belief suggests that such a one-algorithm-fits-all approach cannot work, because this algorithm cannot exploit model specific structure. At least, a generic algorithm cannot be efficient and robust on a wide variety of problems.